July 18, 2011S. Timm FermiCloud Enabling Scientific Computing with Integrated Private Cloud Infrastructures Steven Timm.

Slides:



Advertisements
Similar presentations
Fermilab, the Grid & Cloud Computing Department and the KISTI Collaboration GSDC - KISTI Workshop Jangsan-ri, Nov 4, 2011 Gabriele Garzoglio Grid & Cloud.
Advertisements

Tunis, Tunisia, 28 April 2014 Business Values of Virtualization Mounir Ferjani, Senior Product Manager, Huawei Technologies 2.
Linux Clustering A way to supercomputing. What is Cluster? A group of individual computers bundled together using hardware and software in order to make.
Site Report US CMS T2 Workshop Samir Cury on behalf of T2_BR_UERJ Team.
Introduction to DoC Private Cloud
European Organization for Nuclear Research Virtualization Review and Discussion Omer Khalid 17 th June 2010.
Virtualization for Cloud Computing
Idle virtual machine detection in FermiCloud Giovanni Franzini September 21, 2012 Scientific Computing Division Grid and Cloud Computing Department.
Virtual Network Servers. What is a Server? 1. A software application that provides a specific one or more services to other computers  Example: Apache.
CERN IT Department CH-1211 Genève 23 Switzerland t Next generation of virtual infrastructure with Hyper-V Michal Kwiatek, Juraj Sucik, Rafal.
FermiCloud K. Chadwick, T. Hesselroth, F. Lowe, S. Timm, D. R. Yocum Grid And Cloud Computing Department Fermilab ISGC2011 Work supported by the U.S. Department.
Windows Server MIS 424 Professor Sandvig. Overview Role of servers Performance Requirements Server Hardware Software Windows Server IIS.
Copyright © 2010 Platform Computing Corporation. All Rights Reserved.1 The CERN Cloud Computing Project William Lu, Ph.D. Platform Computing.
OOI CI R2 Life Cycle Objectives Review Aug 30 - Sep Ocean Observatories Initiative OOI CI Release 2 Life Cycle Objectives Review CyberPoPs & Network.
ICT Day Term 4,  Virtualisation is growing in usage.  Current CPU’s are designed to support Virtualisation.  Businesses are looking at virtualisation.
SAIGONTECH COPPERATIVE EDUCATION NETWORKING Spring 2009 Seminar #1 VIRTUALIZATION EVERYWHERE.
Virtualization. Virtualization  In computing, virtualization is a broad term that refers to the abstraction of computer resources  It is "a technique.
+ CS 325: CS Hardware and Software Organization and Architecture Cloud Architectures.
Appendix B Planning a Virtualization Strategy for Exchange Server 2010.
M.A.Doman Short video intro Model for enabling the delivery of computing as a SERVICE.
1 Evolution of OSG to support virtualization and multi-core applications (Perspective of a Condor Guy) Dan Bradley University of Wisconsin Workshop on.
Virtualization within FermiGrid Keith Chadwick Work supported by the U.S. Department of Energy under contract No. DE-AC02-07CH11359.
RAL Site Report Andrew Sansum e-Science Centre, CCLRC-RAL HEPiX May 2004.
The Grid & Cloud Computing Department at Fermilab and the KISTI Collaboration Meeting with KISTI Nov 1, 2011 Gabriele Garzoglio Grid & Cloud Computing.
CERN Database Services for the LHC Computing Grid Maria Girone, CERN.
CLOUD COMPUTING. What is cloud computing ? History Virtualization Cloud Computing hardware Cloud Computing services Cloud Architecture Advantages & Disadvantages.
FermiCloud: Enabling Scientific Workflows with Federation and Interoperability Steven C. Timm FermiCloud Project Lead Grid & Cloud Computing Department.
Globus and PlanetLab Resource Management Solutions Compared M. Ripeanu, M. Bowman, J. Chase, I. Foster, M. Milenkovic Presented by Dionysis Logothetis.
An Introduction to Campus Grids 19-Apr-2010 Keith Chadwick & Steve Timm.
CERN IT Department CH-1211 Genève 23 Switzerland t Next generation of virtual infrastructure with Hyper-V Juraj Sucik, Michal Kwiatek, Rafal.
FermiCloud Project Overview and Progress Keith Chadwick Grid & Cloud Computing Department Head Fermilab Work supported by the U.S. Department of Energy.
Auxiliary services Web page Secrets repository RSV Nagios Monitoring Ganglia NIS server Syslog Forward FermiCloud: A private cloud to support Fermilab.
Unit 2 VIRTUALISATION. Unit 2 - Syllabus Basics of Virtualization Types of Virtualization Implementation Levels of Virtualization Virtualization Structures.
FermiCloud Status Report Fall 2010 Keith Chadwick Grid & Cloud Computing Department Head Fermilab Work supported by the U.S. Department.
Authentication, Authorization, and Contextualization in FermiCloud S. Timm, D. Yocum, F. Lowe, K. Chadwick, G. Garzoglio, D. Strain, D. Dykstra, T. Hesselroth.
Claudio Grandi INFN Bologna Virtual Pools for Interactive Analysis and Software Development through an Integrated Cloud Environment Claudio Grandi (INFN.
WP5 – Infrastructure Operations Test and Production Infrastructures StratusLab kick-off meeting June 2010, Orsay, France GRNET.
Building on virtualization capabilities for ExTENCI Carol Song and Preston Smith Rosen Center for Advanced Computing Purdue University ExTENCI Kickoff.
GPCF* Update Present status as a series of questions / answers related to decisions made / yet to be made * General Physics Computing Facility (GPCF) is.
FermiCloud Review Response to Questions Keith Chadwick Steve Timm Gabriele Garzoglio Work supported by the U.S. Department of Energy under contract No.
RAL Site Report HEP SYSMAN June 2016 – RAL Gareth Smith, STFC-RAL With thanks to Martin Bly, STFC-RAL.
FermiGrid The Fermilab Campus Grid 28-Oct-2010 Keith Chadwick Work supported by the U.S. Department of Energy under contract No. DE-AC02-07CH11359.
Virtualization within FermiGrid Keith Chadwick Work supported by the U.S. Department of Energy under contract No. DE-AC02-07CH11359.
CLOUD COMPUTING Presented to Graduate Students Mechanical Engineering Dr. John P. Abraham Professor, Computer Engineering UTPA.
CASTOR: possible evolution into the LHC era
Bob Jones EGEE Technical Director
New Paradigms: Clouds, Virtualization and Co.
Virtualization for Cloud Computing
Video Security Design Workshop:
DEPARTMENT OF COMPUTER SCIENCE AND ENGINEERING CLOUD COMPUTING
C Loomis (CNRS/LAL) and V. Floros (GRNET)
Blueprint of Persistent Infrastructure as a Service
Dag Toppe Larsen UiB/CERN CERN,
Dag Toppe Larsen UiB/CERN CERN,
Management of Virtual Machines in Grids Infrastructures
Welcome! Thank you for joining us. We’ll get started in a few minutes.
HPEiX Spring RAL Site Report
Virtualization in the gLite Grid Middleware software process
Customer Profile (Target)
WLCG Collaboration Workshop;
Management of Virtual Machines in Grids Infrastructures
OpenStack Ani Bicaku 18/04/ © (SG)² Konsortium.
Dr. John P. Abraham Professor, Computer Engineering UTPA
Public vs Private Cloud Usage Costs:
Virtualization Techniques
Outline Virtualization Cloud Computing Microsoft Azure Platform
Sky Computing on FutureGrid and Grid’5000
HEPiX October 2009 Keith Chadwick.
Virtualization Dr. S. R. Ahmed.
Sky Computing on FutureGrid and Grid’5000
Presentation transcript:

July 18, 2011S. Timm FermiCloud Enabling Scientific Computing with Integrated Private Cloud Infrastructures Steven Timm Fermilab Grid and Cloud Computing Dept. for FermiCloud Project K. Chadwick, F. Lowe, D. Yocum, T. Hesselroth, G. Garzoglio, P. Mhashilkar, T. Levshina SAUCG workshop July 18, 2011 Work supported by the US Dept. of Energy under contract DE-AC02-07CH11359

July 18, 2011S. Timm What FermiCloud is ● Infrastructure-as-a-service private cloud for Fermilab Scientific Program. ● Integrated into Fermilab site security structure. ● Virtual machines have full access to existing Fermilab network and mass storage devices. ● Scientific stakeholders get on-demand access to virtual machines without sysadmin intervention. ● Virtual machines created by users and destroyed or suspended when no longer needed. ● Testbed for developers and integrators to evaluate new grid and storage applications on behalf of scientific stakeholders. ● Ongoing project to build and expand the facility – Technology evaluation, requirements, deployment. – Scalability, monitoring, performance improvement. – High availability and reliability.

July 18, 2011S. Timm Virtualization at Fermilab ● FermiGrid Services – Highly Available provisioned virtual services – SLF5+Xen ● General Physics Compute Facility – Deployment of experiment-specific virtual machines for Intensity Frontier experiments – Oracle VM (Commercialized Xen) ● Virtual Services Group – Virtualization of Fermilab business systems using VMWare – Windows

July 18, 2011S. Timm Science Facility Drivers for FermiCloud ● Continue program of virtualizing all scientific servers that can be virtualized. Many experiment servers need minimal CPU, memory but want ports to themselves. ● Improve utilization of power, cooling and employee time (admins and developers) for managing small science servers. ● Had to replace 6 racks of legacy development machines with limited hardware budget and computer room space. ● CERN IT + HEPiX Virtualisation Taskforce program to have uniformly-deployable virtual machines. Expect LHC and future Fermilab experiments will eventually require cloud technology.

July 18, 2011S. Timm Science Stakeholders ● Joint Dark Energy Mission->WFIRST->LSST – Distributed messaging system, testing fault tolerance. ● Fermilab Intensity Frontier – Monitoring Server (MCAS) – GridFTP endpoint server – Experiment-specific storage investigations ● Fermilab D0 Experiment – Job Forwarding Server ● Extenci project (Cloud activities, LHC) – Distributed storage on WAN. ● GEANT4 – Validation server ● Scientific middleware development hosting – Host developers and integrators of OSG middleware.

July 18, 2011S. Timm Hardware ● 2x Quad Core Intel Xeon E5640 CPU ● 2 SAS 15K RPM system disk 300GB ● 6x 2TB SATA disk ● LSI 1078 RAID controller ● Infiniband card ● 24GB RAM ● 23 machines total ● Arrived June 2010 ● +25TB Bluearc NAS disk

July 18, 2011S. Timm Software Technologies ● OS: Scientific Linux 5, 6 ● Hypervisor: KVM ● Cloud Management: OpenNebula ● Modifications to OpenNebula CLI, Query API, GUI to use X.509 authentication to launch virtual machines. (See my talk at OGF32.) ● Secure credential store—all security secrets loaded at runtime only. ● Site-wide patching and vulnerability scanning facilities.

July 18, 2011S. Timm Current Technology Investigations ● Testing storage services with real neutrino experiment codes, identify NFS alternatives. ● Using Infiniband interface to create sandbox for MPI applications. ● Batch queue look-ahead to create worker node VM's on demand. ● Submission of multiple worker node VM's, grid cluster in the cloud. ● Idle VM detection and suspension, backfill with worker node VM's. ● Leverage site “network jail” for new virtual machines. ● IPv6 support.

July 18, 2011S. Timm High Availability and Service Levels ● Add SAN for live migration and large datablock capacity. ● Split FermiCloud between two buildings ● Mirror storage between two buildings ● Set up high-availability procedures for failover of cloud controller and migration of virtual machines. ● Offer three service levels – High availability 24x7 – Regular virtual machine – Opportunistic (spot market) can be pre-empted anytime. ● Stakeholders billed for usage according to an economic model, analogous to existing tape robot facility.

July 18, 2011S. Timm Ongoing software development ● Accounting and billing—Cloud accounting add-ons to Gratia accounting project. ● Monitoring—How many machines are running, who is running them, is everything up that should be up? ● Authorization—Apply well-tested and interoperable grid authorization tools to cloud authorization as well. ● All of above in collaboration with other projects and standards bodies.

July 18, 2011S. Timm Using Virtualization to Enable Science ● New interactive science applications that require ongoing interaction or unique network topologies and don't fit grid batch processing paradigm. ● Complicated software stacks where grid distribution has been difficult or impossible. ● Legacy experiments which require specific OS and library combinations. ● Extra compute capacity on demand for experiments that need it. ● Virtualization used on 32-core+ worker nodes to – Pin applications to appropriate CPU-memory combinations for better performance – Sandbox applications to keep one rogue job from killing the other 31. – Memory segments can grow or be shared as needed.

July 18, 2011S. Timm Conclusions ● FermiCloud has successfully deployed a wide range of servers for the scientific program. ● FermiCloud has been a testbed for several evaluations of storage and middleware that benefit the scientific program. ● FermiCloud has already provided significant power and cooling savings, and significant convenience benefits to scientific stakeholders ● Now integrating our work with other internal Fermilab virtualization activities and external projects. ● We welcome interest from new users, stakeholders, and other cloud-based projects.